Abstract
Recent advances such as OpenAI-o1 and DeepSeek R1 have demonstrated the potential of Reinforcement Learning (RL) to enhance reasoning abilities in Large Language Models (LLMs). While open-source replication efforts have primarily focused on mathematical and coding domains, methods and resources for developing general reasoning capabilities remain underexplored. This gap is partly due to the challenge of collecting diverse and verifiable reasoning data suitable for RL. We hypothesize that logical reasoning is critical for developing general reasoning capabilities, as logic forms a fundamental building block of reasoning. In this work, we present SynLogic, a data synthesis framework and dataset that generates diverse logical reasoning data at scale, encompassing 35 diverse logical reasoning tasks. The SynLogic approach enables controlled synthesis of data with adjustable difficulty and quantity. Importantly, all examples can be verified by simple rules, making them ideally suited for RL with verifiable rewards. In our experiments, we validate the effectiveness of RL training on the SynLogic dataset based on 7B and 32B models. SynLogic leads to state-of-the-art logical reasoning performance among open-source datasets, surpassing DeepSeek-R1-Distill-Qwen-32B by 6 points on BBEH. Furthermore, mixing SynLogic data with mathematical and coding tasks improves the training efficiency of these domains and significantly enhances reasoning generalization. Notably, our mixed training model outperforms DeepSeek-R1-Zero-Qwen-32B across multiple benchmarks. These findings position SynLogic as a valuable resource for advancing the broader reasoning capabilities of LLMs. We open-source both the data synthesis pipeline and the SynLogic dataset at https://github.com/MiniMax-AI/SynLogic.
Abstract
Recent advances such as OpenAI-o1 and DeepSeek R1 have demonstrated the potential of Reinforcement Learning (RL) to enhance reasoning abilities in Large Language Models (LLMs). While open-source replication efforts have primarily focused on mathematical and coding domains, methods and resources for developing general reasoning capabilities remain underexplored. This gap is partly due to the challenge of collecting diverse and verifiable reasoning data suitable for RL. We hypothesize that logical reasoning is critical for developing general reasoning capabilities, as logic forms a fundamental building block of reasoning. In this work, we present SynLogic, a data synthesis framework and dataset that generates diverse logical reasoning data at scale, encompassing 35 diverse logical reasoning tasks. The SynLogic approach enables controlled synthesis of data with adjustable difficulty and quantity. Importantly, all examples can be verified by simple rules, making them ideally suited for RL with verifiable rewards. In our experiments, we validate the effectiveness of RL training on the SynLogic dataset based on 7B and 32B models. SynLogic leads to state-of-the-art logical reasoning performance among open-source datasets, surpassing DeepSeek-R1-Distill-Qwen-32B by 6 points on BBEH. Furthermore, mixing SynLogic data with mathematical and coding tasks improves the training efficiency of these domains and significantly enhances reasoning generalization. Notably, our mixed training model outperforms DeepSeek-R1-Zero-Qwen-32B across multiple benchmarks. These findings position SynLogic as a valuable resource for advancing the broader reasoning capabilities of LLMs. We open-source both the data synthesis pipeline and the SynLogic dataset at https://github.com/MiniMax-AI/SynLogic.
![[Uncaptioned image]](x1.png)
1 Introduction

The success of Deepseek R1 (DeepSeek-AI et al., 2025) and OpenAI-o1 (Jaech et al., 2024) demonstrates the great potential of post-training in advancing strong reasoning capabilities. These works reveal that the core methodology behind these advancements is reinforcement learning with verifiable rewards (RLVR), inspiring numerous replication efforts focused on RL training. However, most of these works have concentrated on the mathematics and coding domains, primarily because it is straightforward to design binary reward rules in these areas (Zeng et al., 2025b, a; Yu et al., 2025; Hu et al., 2025; Zhang et al., 2025). To foster more general and comprehensive reasoning abilities, it is essential to utilize diverse tasks and examples with verifiable rewards. In this work, we concentrate on logical reasoning as a promising domain for this objective, hypothesizing that logical reasoning serves as a fundamental building block for developing general reasoning skills. Although prior work has explored RL training in the context of logic tasks (Xie et al., 2025b; Pan et al., 2025), these efforts have typically focused on a single task, leaving the potential of broader and more diverse synthetic logic datasets largely underexplored.
Synthetic logic data presents distinct advantages and challenges. Its synthetic nature allows for unlimited data generation with controllable difficulty levels, enabling the creation of increasingly challenging samples. Additionally, the intrinsic properties of some logic tasks like Sudoku often require trial and backtracking in the reasoning process, which closely relates to the “aha moments” (DeepSeek-AI et al., 2025) in problem-solving. Therefore, the primary advantages of synthetic logic data for RLVR lie in its scalability and inherent characteristics that align well with complex reasoning processes. The main challenge, however, is the complexity of generating and designing specific rules for different logic tasks, as tasks like Game of 24 and Sudoku each require distinct verifiers.
Recent works primarily focus on logic evaluation (Suzgun et al., 2022; Ma et al., 2024; Kazemi et al., 2025), but lack high-quality accessible logical reasoning training data. In this work, to address the gap in comprehensive logic tasks, we present SynLogic: a logical reasoning data synthesis framework and a comprehensive synthetic logic dataset containing 35 tasks, including typical logical tasks such as Sudoku, Game of 24, and Cipher. For each task, we develop task-specific generation code paired with a corresponding rule-based verifier, allowing for fine-grained difficulty control through adjustable generation hyperparameters.
To validate the effectiveness of reinforcement learning on the SynLogic data, we run RL training on it with the GRPO algorithm (Shao et al., 2024) and implement binary outcome rewards determined by each task’s verification rules. By adapting recent GRPO training techniques introduced in DAPO (Yu et al., 2025), we successfully train Qwen2.5 Base models (Yang et al., 2024) on the SynLogic data in a zero RL training setting, achieving progressively longer COT responses and observing the emergence of reflection behaviors. Starting from Qwen2.5-7B-Base and Qwen2.5-32B-Base foundations, our models achieve over 8 absolute percentage points improvement on the logic benchmark KOR-Bench (Ma et al., 2024) compared to their instruction models. Notably, our 32B model surpasses DeepSeek-R1-Distill-Qwen-32B on BBEH (Kazemi et al., 2025) tasks by 5 absolute points, establishing SynLogic as the state-of-the-art open-source dataset for logical reasoning to date. Additionally, both models demonstrate strong generalization to unseen mathematics domains.
Furthermore, we explore mixing the SynLogic data with mathematics or coding data for RL training. Surprisingly, conducting the mixed training on Qwen2.5-7B-Base model (Yang et al., 2024), incorporating SynLogic data improves training efficiency for developing mathematical and coding skills. For mathematics, mixed training maintains similar mathematics performance under the same number of training steps, which consume fewer math training samples. Simultaneously, mixed training achieves much higher performance on logic tasks. A similar trend is observed when mixing SynLogic with coding data, further demonstrating the complementary benefits of logical reasoning training. Finally, we conduct large-scale mixed training on the Qwen2.5-32B-Base model to enhance the capability of Zero-RL training. Our mixed training achieves superior performance on multiple benchmarks compared to the DeepSeek-R1-Zero-Qwen-32B model, consistently outperforming or matching it on BBEH (Kazemi et al., 2025), KOR-Bench (Ma et al., 2024), LiveCodeBench (Jain et al., 2025), and GPQA-Diamond (Rein et al., 2024), validating the generalization benefits provided by the inclusion of logical reasoning data.
2 SynLogic: Synthesizing Logical Reasoning Data at Scale
Dataset | Tasks | Trainable | Adjustable Difficulty |
BBH (Suzgun et al., 2022) | 23 | ✗ | ✗ |
Zebra Logic (Lin et al., 2024) | 1 | ✗ | ✓ |
KOR-Bench (Ma et al., 2024) | 5* | ✗ | ✗ |
K&K (Xie et al., 2025a) | 1 | ✓ | ✓ |
BBEH (Kazemi et al., 2025) | 23 | ✗ | ✗ |
SynLogic | 35 | ✓ | ✓ |
2.1 Background
Logical reasoning has long been a crucial indicator of model intelligence (ai2, 2019; Suzgun et al., 2022), valued for both its significance and synthetic accessibility. With the advancement of reasoning capabilities in Large Language Models (LLMs), researchers have developed increasingly challenging benchmarks to evaluate logical reasoning abilities (Kazemi et al., 2025; Ma et al., 2024). However, as illustrated in Table 1, existing benchmarks either lack training support or are limited to a small number of tasks. Synthetic logic data serves as an important source of verifiable data and offers straightforward control over task difficulty, presenting the potential for developing scalable stronger models by training on it. Consequently, comprehensive synthetic logic datasets are essential for developing general strong reasoning models (Seed et al., 2025).
2.2 The Data Synthesis Framework
To synthesize large-scale, diverse synthetic data, we develop a comprehensive data synthesis framework encompassing 35 tasks. While the benchmarks in Table 1 include a wide variety of tasks, a significant challenge we faced is that nearly all evaluation benchmarks do not open-source their data generation methods. This hinders us from building training data for these logic tasks directly. Therefore, we develop these tasks independently, building the SynLogic framework, illustrated in Figure 1. The framework consists of the following key components:
-
1.
Task Selection We select a diverse set of logic tasks that require non-trivial reasoning, drawing from two carefully curated categories of data sources: (1) widely recognized puzzle problems from logic communities, such as the game of 24, Sudoku, and cryptarithms. Many of these puzzles have been previously highlighted in works like (Kurtic et al., 2024; Li et al., 2024; Ma et al., 2024). (2) Logic tasks featured in established evaluation benchmarks, including BBH (Suzgun et al., 2022) and BBEH (Kazemi et al., 2025). Detailed descriptions and sources for all 35 tasks can be found in the Appendix A.1.
-
2.
Parameter Identification For each task, we identify key parameters that control difficulty (e.g., grid size in Sudoku, or missing numbers in Math Path). These parameters form the basis for scalable and adjustable difficulty data synthesis.
-
3.
Logic Instance Generation We formalize the task-specific rules into code by manually implementing rule-based logic generators for each task. These generators are designed to encode the specific constraints and rules of the logic problems, ensuring that the generated instances adhere to the intended task structure (e.g., enforcing the unique digits rule in Sudoku). This rule-based approach allows us to efficiently produce large quantities of data and to cover a broad spectrum of difficulty levels by adjusting the difficulty related parameters. All generated instances undergo automated checks for correctness and solvability.
-
4.
Appropriate Difficulty Control To ensure that the generated data is both challenging and learnable, we carefully adjust difficulty-related parameters during data generation. We use strong reasoning models, DeepSeek R1 (DeepSeek-AI et al., 2025) and OpenAI-o3-mini to set an upper bound on difficulty: the highest difficulty parameters for which R1 or o3-mini can solve samples with a pass@10 greater than zero, representing the limit of these models’ solvability. This approach prevents the inclusion of instances that are too difficult. Similarly, we use chat models to determine the lower bound of difficulty: the lowest difficulty parameters for which the models achieve a pass rate between 0 and 0.5. This dual-bound approach ensures that the dataset includes a balanced range of samples, maintaining an appropriate level of complexity and learnability.
-
5.
Prompt Formalization To facilitate training and evaluation with LLMs, we convert abstract logic instances into natural language prompts using task-specific prompt templates. This step ensures that each instance is accessible to both humans and language models.
-
6.
Verification Suite For every task, we implement a dedicated verifier that can automatically check the correctness of model outputs, supporting both training supervision and automatic evaluation.
A key innovation in our approach is the development of customized difficulty control mechanisms for each task type. Unlike existing benchmarks that often provide fixed-difficulty evaluation data, our system allows precise calibration of problem complexity through task-specific parameters, such as grid size in Sudoku. This difficulty-tuning capability enables the creation of different difficulty level data, presenting the potential of progressively challenging training curricula. We overcame significant challenges in implementing these controls, as many evaluation benchmarks do not open-source their data generation methods. At last, most tasks in SynLogic are designed with: (1) a data generation code capable of producing varied instances, (2) a corresponding verification rule for evaluating solution correctness, and (3) configurable difficulty parameters to enable controlled difficulty of generated data. We independently develop and generate data for 33 tasks in our dataset, while only the data of 2 tasks (Zebra Puzzle (Lin et al., 2024) and ARC-AGI (Chollet, 2019)) are directly adopted from existing open source resources.
2.2.1 Risk of Data Contamination
Although several tasks overlap between our selected tasks and current benchmarks, such as KOR-Bench and BBEH, the synthetic nature of our data, combined with the large synthesis space, makes the probability of generating data identical to benchmark test samples very low – we have verified that there are no identical samples between our generated datasets and the benchmark test sets.
2.3 The SynLogic Datasets
We synthesized our dataset with controlled difficulty parameters for each task, carefully balancing challenge and learnability to ensure the success of our subsequent experiments §3. To accommodate different model capacities, we developed two distinct versions of our dataset: SynLogic-Hard for Qwen2.5-32B training and SynLogic-Easy for Qwen2.5-7B training. SynLogic-Hard presents more complex challenges with its broader difficulty level for each tasks with difficulty upper bound described in § 2.2. For SynLogic-Easy, we systematically lower difficulty parameters across all tasks to create easier versions. Despite these adjustments, eight tasks still remain beyond the learning capacity of the 7B model with zero training accuracy after RL, thus we removed them from this easy version, where the details about the removed tasks are provided in Appendix A.2. Finally, we synthesized 33k SynLogic-Hard samples and 16k SynLogic-Easy samples used in subsequent experiments for training, along with 10 validation samples per task, separately for the Easy and Hard validation splits.
2.3.1 Difficulty Analysis
To assess the difficulty of the synthetic data, we conduct an evaluation on the validation splits, assessing model performance using both avg@8 (average pass rate with eight attempts) and pass@8 (success within eight attempts) metrics. The results, illustrated in Figure 2, confirm the appropriate difficulty levels for each model scale, demonstrating that our datasets provide suitable training challenges across different model capacities.


3 Reinforcement Learning on SynLogic
Model | Logic Benchmarks | Mathematical Benchmarks | |||||
SynLogic-Val | KOR-Bench | BBH | BBEH | AIME 2024 | MATH 500 | AMC 2023 | |
Qwen2.5-7B-Base | 2.8 | 11.6 | 45.2 | 3.8 | 0.3 | 64.6 | 30.0 |
Qwen2.5-7B-Instruct | 9.0 | 38.6 | 62.7 | 12.4 | 6.3 | 76.4 | 52.5 |
SynLogic-7B | 44.4 | 48.1 | 66.5 | 8.0 | 10.0 | 71.8 | 55.0 |
Qwen2.5-32B-Base | 1.6 | 10.9 | 58.4 | 3.3 | 4.5 | 68.6 | 45.0 |
Qwen2.5-32B-Instruct | 12.0 | 54.7 | 84.5 | 17.5 | 10.0 | 82.2 | 57.5 |
R1-Distill-Qwen-32B | 33.0 | 66.6 | 88.3 | 19.2 | 72.6 | 94.3 | 85.0 |
SynLogic-32B | 52.9 | 62.2 | 85.8 | 25.5 | 19.6 | 82.0 | 57.5 |
Reinforcement learning with verifiable rewards (RLVR) has emerged as a highly effective approach for enhancing reasoning capabilities in large language models (Zeng et al., 2025b; Yu et al., 2025; Hu et al., 2025; Zhang et al., 2025). Building on these advances, our experimental framework also focuses on applying reinforcement learning techniques to the SynLogic dataset, leveraging the verifiable nature of logical reasoning tasks. In this section, we validate the effectiveness of reinforcement learning training on the SynLogic dataset using Qwen2.5-7B-Base and Qwen2.5-32B-Base models.
3.1 Setup Details
3.1.1 Training Template
Following the DAPO training prompt (Yu et al., 2025), we modify and design the training prompt template for logic training as shown in Figure 3:
3.1.2 Reward Design
Our reward function employs a binary scoring mechanism that evaluates both format adherence and answer correctness. Specifically, we assign a reward of 1 only when a model-generated response satisfies two criteria: (1) it correctly follows the designated format by including both <think> </think> and <answer> </answer> tags, and (2) the final answer provided is correct. Responses that either deviate from the required format or contain incorrect answers receive a reward of 0.
(1) |
where evaluates whether response includes both the required <think> </think> and <answer> </answer> tags, and determines whether the answer provided is correct verified by its task’s verification rule.
3.1.3 Training Details
For our experiments, we synthesized approximately 16k SynLogic-Easy and 33k SynLogic-Hard instances to train the Qwen2.5-7B-Base and Qwen2.5-32B-Base models with DAPO, respectively, as described in §2.3. During training, we employed a prompt batch size of 128, generated 16 rollouts per prompt, and set maximum rollout lengths of 16,384 tokens for the 7B model and 28,672 tokens for the 32B model. We configured the clip high parameter at 0.28. Additional training hyperparameters and implementation details are provided in Appendix B.1.2.
3.1.4 Evaluation Details
Our evaluation strategy encompasses two distinct benchmark categories. For assessing logical reasoning capabilities, we employ the validation splits of SynLogic alongside established benchmarks including Knowledge-Orthogonal Reasoning (KOR-Bench) (Ma et al., 2024), BBH (Suzgun et al., 2022), and the substantially more challenging BBEH (Kazemi et al., 2025). To investigate cross-domain generalization effects, we incorporate mathematics evaluations on MATH 500 (Hendrycks et al., 2021), AMC 2023, and AIME 2024. All evaluations are conducted in a zero-shot setting, with avg@8 metrics computed for AIME 2024 and SynLogic-Val to mitigate variance.
3.2 Results


3.2.1 Substantial Improvements in Logical Reasoning
The evaluation results presented in Table 2 demonstrate significant improvements across logical reasoning tasks. Beyond the notable gains on SynLogic’s validation split, our models demonstrate enhanced performance across multiple logical benchmarks, leading state-of-the-art results among open-source datasets. Our 7B model achieves 48.1% on KOR-Bench (Ma et al., 2024), outperforming Qwen2.5-7B-Instruct by nearly 10 absolute percentage points. Similarly, our 32B model surpasses Qwen2.5-32B-Instruct by 7 percentage points on KOR-Bench. Notably, our 32B model exceeds R1-Distill-Qwen32B (DeepSeek-AI et al., 2025) by 6 percentage points on the challenging BBEH benchmark (Kazemi et al., 2025), showcasing the effectiveness of the SynLogic dataset in driving state-of-the-art logical reasoning performance.
3.2.2 Generalization to Mathematical Domains
Our experimental results demonstrate significant generalization capabilities to mathematical domains, as shown in Table 2. Despite being primarily trained for logical reasoning, SynLogic models exhibit strong performance across mathematical benchmarks over their base models. SynLogic-7B achieves 10.0% on AIME 2024, a nearly 10 absolute point improvement compared to Qwen2.5-7B-Base (0.3%), 71.8% on MATH 500, a 7.2 absolute point gain, and 55.0% on AMC 2023, a 25-point increase. More remarkably, SynLogic-32B achieves 19.6% on AIME 2024, a 4.4x improvement over Qwen2.5-32B-Base (4.5%), while its performance on MATH 500 (82.0%) and AMC 2023 (57.5%) shows substantial gains of 13.4 and 12.5 points, respectively. Without mathematics training data, our models nearly match or surpass the instruction models, suggesting that enhancements in logical reasoning capabilities transfer effectively to mathematical problem-solving. This aligns with the observation in Logic-RL (Xie et al., 2025b), highlighting the fundamental connection between logical and mathematical reasoning skills.
3.2.3 Increased Chain-of-Thought Length
As shown in Figure 4, recording the response length and reflection ratio during the training process reveals that training on SynLogic data leads to stable increases in response length for both models. The 7B model reaches an average of approximately 2500 tokens, while the 32B model achieves around 4000 tokens. Additionally, the increasing reflection ratios also indicate the emergence of cognitive behaviors during training (Gandhi et al., 2025). Both the extended response lengths and increased prevalence of reflection tokens suggest that synthetic logic reasoning tasks inherently align with the long-thinking paradigm of LLM. Detailed training accuracy results are provided in Appendix B.1.3.
4 Scaling RL Training with Diverse Verifiable Reasoning Data
Having verified the success of reinforcement training on SynLogic alone, we now leverage verifiable reasoning data from math, coding, and logical reasoning domains, scaling RL training with diverse verifiable reasoning data. Concretely, we mix SynLogic with mathematical or/and coding training data, and perform RL training on the mixed datasets. We will study how combining logical reasoning with code and mathematical data influences training efficiency on 7B models and enhances the Zero-RL capabilities for 32B models.
4.1 Setup Details
For mathematical training data, we directly utilize the 17k samples provided in DAPO (Yu et al., 2025). For coding data, we assembled approximately 9k samples from various online coding platforms such as Codeforces. We adapted a similar prompt template to that used in our SynLogic training; the detailed templates are presented in Appendix B.1. We maintained the same reward design approach as described in §3.1.2. Specifically, for coding tasks, the reward is set as 1 if the output is correctly formatted and all test cases pass; otherwise, the reward is 0.






4.2 Mixing SynLogic with Math or Code Data: A Pilot Ablation Study
4.2.1 Mixed Training with Math
We sample approximately 17k samples from SynLogic-Easy and combine them with 17k math data for training on the Qwen2.5-7B-Base model. For controlled comparison, we also conduct reinforcement learning using exclusively math data. Both experimental configurations maintain identical hyperparameters, optimization settings, and computational resources to ensure fair evaluation. Figure 5 presents a comparison of training dynamics. Running for the same number of training steps, mixed training (Logic+Math) achieves comparable performance to math-only training on average across three mathematical benchmarks (Figure 5b), while consuming fewer math samples. Under the same volume of processed math data, mixed training achieves higher accuracy (Figure 5c). Moreover, mixed training steadily improves logical reasoning, as reflected in rising KOR-Bench scores (Figure 5a), which are nearly 10 absolute percentage points higher than those achieved with math-only training. These results suggest that mixed training facilitates more efficient optimization, potentially due to shared abstract reasoning mechanisms across domains.
4.2.2 Mixed Training with Code
Following a similar methodology, we sample approximately 9k samples from SynLogic-Easy and combine them with 9K code samples to train the Qwen2.5-7B-Base model. As a control, we conduct parallel training using exclusively coding data. Both training configurations maintain identical parameters to ensure a fair comparison. To measure the coding ability, we include the validation split of our coding data and the LiveCodeBench (Jain et al., 2025) (the same version as used in DeepSeek’s report (DeepSeek-AI et al., 2025)) for evaluation. As shown in Figure 6, we observe a similar phenomenon of more efficient training dynamics when mixing code with SynLogic-Easy. Models trained on Logic+Coding data achieve higher performance on coding benchmarks than code-only training when consuming the same volume of coding data. Simultaneously, mixed training improves logical reasoning, as evidenced by 10 absolute points better KOR-Bench scores (Figure 6a). These findings reinforce the complementary nature of logical reasoning in enhancing domain-specific capabilities.
4.3 32B Zero-RL Training with Diverse Reasoning Data
Model | BBEH | KOR-Bench | LCB | AIME 2024 | GPQA Diamond |
DeepSeek-R1-Distill-Qwen-32B | 19.2 | 66.6 | 57.2 | 72.6 | 63.1 |
DeepSeek-R1-Zero-Qwen-32B | - | - | 40.2 | 47.0 | 55.0 |
Zero-Mix-2 (Math+Coding) | 18.5 | 58.6 | 39.5 | 34.5 | 55.2 |
Zero-Mix-3 (SynLogic+Math+Coding) | 28.6 | 65.0 | 40.7 | 35.8 | 57.5 |
Building on the previous observation, here we scale up the diverse, verifiable training data by mixing math, coding, and SynLogic datasets, and perform RL training on the Qwen2.5-32B-Base model. Specifically, we use a mix of 35k mathematical samples, 9k coding samples, and 17k SynLogic samples for training. We term this training configuration as Zero-Mix-3. We additionally conduct a Zero-Mix-2 setting that only mixes coding and mathematical data, serving as an ablation baseline to study the effect of SynLogic in such a scalable setting. Related to our Zero-Mix-2 setup, Zhang et al. (2025) recently demonstrated that combining mathematical data with coding data is able to facilitate coding learning. Here we further scale up this trend by including our proposed SynLogic dataset. To evaluate generalization, we include an out-of-domain benchmark, GPQA Diamond (Rein et al., 2024), to study how the addition of SynLogic impacts broader reasoning capabilities. Both Zero-Mix-3 and Zero-Mix-2 configurations are run for the same number of training steps to ensure a fair and controlled comparison.
4.3.1 Results
As shown in Table 3, the Zero-RL training in Zero-Mix-3 (SynLogic+Math+Coding) achieves superior performance across multiple evaluations. On logic benchmarks, Zero-Mix-3 nearly matches the performance of DeepSeek-R1-Distill-Qwen-32B on KOR-Bench and surpasses it by 8 points on BBEH. Notably, Zero-Mix-3 also matches DeepSeek-R1-Zero-Qwen-32B on the coding benchmark LiveCodeBench (Jain et al., 2025) and outperforms it on GPQA-Diamond. Compared to the Zero-Mix-2 (Math+Coding) experiment, Zero-Mix-3 consistently delivers higher performance across all benchmarks. Specifically, Zero-Mix-3 shows a significant improvement of over 10 points on BBEH, 6 points on KOR-Bench, and over 2 points on the out-of-domain benchmark GPQA Diamond. These results strongly validate the significant generalization benefits provided by the inclusion of SynLogic.
5 Conclusion
We present SynLogic: a data synthesis framework and a comprehensive synthetic logic dataset with 35 diverse tasks, addressing the lack of high-quality logic training data. Using SynLogic, we trained Qwen2.5 models with the GRPO algorithm, achieving significant gains on logic benchmarks like KOR-Bench and strong generalization to unseen mathematical tasks. Notably, our 32B model outperformed DeepSeek-R1-Distill-Qwen-32B on BBEH. Mixed training with SynLogic further improved training efficiency and performance, showcasing the complementary benefits of logical reasoning across domains. We hope SynLogic inspires broader exploration of synthetic datasets and logical reasoning to develop stronger reasoning capability models.
References
- ai2 (2019) Winogrande: An adversarial winograd schema challenge at scale. 2019.
- Chollet (2019) François Chollet. On the measure of intelligence. arXiv preprint arXiv:1911.01547, 2019.
- DeepSeek-AI et al. (2025) DeepSeek-AI, Daya Guo, Dejian Yang, Haowei Zhang, Junxiao Song, Ruoyu Zhang, Runxin Xu, Qihao Zhu, Shirong Ma, Peiyi Wang, Xiao Bi, Xiaokang Zhang, Xingkai Yu, Yu Wu, Z. F. Wu, Zhibin Gou, Zhihong Shao, Zhuoshu Li, Ziyi Gao, Aixin Liu, Bing Xue, Bingxuan Wang, Bochao Wu, Bei Feng, Chengda Lu, Chenggang Zhao, Chengqi Deng, Chenyu Zhang, Chong Ruan, Damai Dai, Deli Chen, Dongjie Ji, Erhang Li, Fangyun Lin, Fucong Dai, Fuli Luo, Guangbo Hao, Guanting Chen, Guowei Li, H. Zhang, Han Bao, Hanwei Xu, Haocheng Wang, Honghui Ding, Huajian Xin, Huazuo Gao, Hui Qu, Hui Li, Jianzhong Guo, Jiashi Li, Jiawei Wang, Jingchang Chen, Jingyang Yuan, Junjie Qiu, Junlong Li, J. L. Cai, Jiaqi Ni, Jian Liang, Jin Chen, Kai Dong, Kai Hu, Kaige Gao, Kang Guan, Kexin Huang, Kuai Yu, Lean Wang, Lecong Zhang, Liang Zhao, Litong Wang, Liyue Zhang, Lei Xu, Leyi Xia, Mingchuan Zhang, Minghua Zhang, Minghui Tang, Meng Li, Miaojun Wang, Mingming Li, Ning Tian, Panpan Huang, Peng Zhang, Qiancheng Wang, Qinyu Chen, Qiushi Du, Ruiqi Ge, Ruisong Zhang, Ruizhe Pan, Runji Wang, R. J. Chen, R. L. Jin, Ruyi Chen, Shanghao Lu, Shangyan Zhou, Shanhuang Chen, Shengfeng Ye, Shiyu Wang, Shuiping Yu, Shunfeng Zhou, Shuting Pan, S. S. Li, Shuang Zhou, Shaoqing Wu, Shengfeng Ye, Tao Yun, Tian Pei, Tianyu Sun, T. Wang, Wangding Zeng, Wanjia Zhao, Wen Liu, Wenfeng Liang, Wenjun Gao, Wenqin Yu, Wentao Zhang, W. L. Xiao, Wei An, Xiaodong Liu, Xiaohan Wang, Xiaokang Chen, Xiaotao Nie, Xin Cheng, Xin Liu, Xin Xie, Xingchao Liu, Xinyu Yang, Xinyuan Li, Xuecheng Su, Xuheng Lin, X. Q. Li, Xiangyue Jin, Xiaojin Shen, Xiaosha Chen, Xiaowen Sun, Xiaoxiang Wang, Xinnan Song, Xinyi Zhou, Xianzu Wang, Xinxia Shan, Y. K. Li, Y. Q. Wang, Y. X. Wei, Yang Zhang, Yanhong Xu, Yao Li, Yao Zhao, Yaofeng Sun, Yaohui Wang, Yi Yu, Yichao Zhang, Yifan Shi, Yiliang Xiong, Ying He, Yishi Piao, Yisong Wang, Yixuan Tan, Yiyang Ma, Yiyuan Liu, Yongqiang Guo, Yuan Ou, Yuduan Wang, Yue Gong, Yuheng Zou, Yujia He, Yunfan Xiong, Yuxiang Luo, Yuxiang You, Yuxuan Liu, Yuyang Zhou, Y. X. Zhu, Yanhong Xu, Yanping Huang, Yaohui Li, Yi Zheng, Yuchen Zhu, Yunxian Ma, Ying Tang, Yukun Zha, Yuting Yan, Z. Z. Ren, Zehui Ren, Zhangli Sha, Zhe Fu, Zhean Xu, Zhenda Xie, Zhengyan Zhang, Zhewen Hao, Zhicheng Ma, Zhigang Yan, Zhiyu Wu, Zihui Gu, Zijia Zhu, Zijun Liu, Zilin Li, Ziwei Xie, Ziyang Song, Zizheng Pan, Zhen Huang, Zhipeng Xu, Zhongyu Zhang, and Zhen Zhang. Deepseek-r1: Incentivizing reasoning capability in llms via reinforcement learning, 2025. URL https://arxiv.org/abs/2501.12948.
- Gandhi et al. (2025) Kanishk Gandhi, Ayush Chakravarthy, Anikait Singh, Nathan Lile, and Noah D. Goodman. Cognitive behaviors that enable self-improving reasoners, or, four habits of highly effective stars, 2025. URL https://arxiv.org/abs/2503.01307.
- Hendrycks et al. (2021) Dan Hendrycks, Collin Burns, Saurav Kadavath, Akul Arora, Steven Basart, Eric Tang, Dawn Song, and Jacob Steinhardt. Measuring mathematical problem solving with the math dataset. arXiv preprint arXiv:2103.03874, 2021.
- Hu et al. (2025) Jingcheng Hu, Yinmin Zhang, Qi Han, Daxin Jiang, Xiangyu Zhang, and Heung-Yeung Shum. Open-reasoner-zero: An open source approach to scaling up reinforcement learning on the base model. arXiv preprint arXiv:2503.24290, 2025.
- Jaech et al. (2024) Aaron Jaech, Adam Kalai, Adam Lerer, Adam Richardson, Ahmed El-Kishky, Aiden Low, Alec Helyar, Aleksander Madry, Alex Beutel, Alex Carney, et al. Openai o1 system card. arXiv preprint arXiv:2412.16720, 2024.
- Jain et al. (2025) Naman Jain, King Han, Alex Gu, Wen-Ding Li, Fanjia Yan, Tianjun Zhang, Sida Wang, Armando Solar-Lezama, Koushik Sen, and Ion Stoica. Livecodebench: Holistic and contamination free evaluation of large language models for code. In The Thirteenth International Conference on Learning Representations, 2025. URL https://openreview.net/forum?id=chfJJYC3iL.
- Kazemi et al. (2025) Mehran Kazemi, Bahare Fatemi, Hritik Bansal, John Palowitch, Chrysovalantis Anastasiou, Sanket Vaibhav Mehta, Lalit K Jain, Virginia Aglietti, Disha Jindal, Peter Chen, et al. Big-bench extra hard. arXiv preprint arXiv:2502.19187, 2025.
- Kurtic et al. (2024) Eldar Kurtic, Amir Moeini, and Dan Alistarh. Mathador-LM: A dynamic benchmark for mathematical reasoning on large language models. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen, editors, Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 17020–17027, Miami, Florida, USA, November 2024. Association for Computational Linguistics. 10.18653/v1/2024.emnlp-main.946. URL https://aclanthology.org/2024.emnlp-main.946/.
- Li et al. (2024) Yinghao Li, Haorui Wang, and Chao Zhang. Assessing logical puzzle solving in large language models: Insights from a minesweeper case study. In Kevin Duh, Helena Gomez, and Steven Bethard, editors, Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers), pages 59–81, Mexico City, Mexico, June 2024. Association for Computational Linguistics. 10.18653/v1/2024.naacl-long.4. URL https://aclanthology.org/2024.naacl-long.4/.
- Lin et al. (2024) Bill Yuchen Lin, Ronan Le Bras, and Yejin Choi. Zebralogic: Benchmarking the logical reasoning ability of language models, 2024. URL https://hf. co/spaces/allenai/ZebraLogicBench-Leaderboard, 2024.
- Ma et al. (2024) Kaijing Ma, Xinrun Du, Yunran Wang, Haoran Zhang, Zhoufutu Wen, Xingwei Qu, Jian Yang, Jiaheng Liu, Minghao Liu, Xiang Yue, et al. Kor-bench: Benchmarking language models on knowledge-orthogonal reasoning tasks. arXiv preprint arXiv:2410.06526, 2024.
- Pan et al. (2025) Jiayi Pan, Junjie Zhang, Xingyao Wang, Lifan Yuan, Hao Peng, and Alane Suhr. Tinyzero. https://github.com/Jiayi-Pan/TinyZero, 2025. Accessed: 2025-01-24.
- Rein et al. (2024) David Rein, Betty Li Hou, Asa Cooper Stickland, Jackson Petty, Richard Yuanzhe Pang, Julien Dirani, Julian Michael, and Samuel R. Bowman. GPQA: A graduate-level google-proof q&a benchmark. In First Conference on Language Modeling, 2024. URL https://openreview.net/forum?id=Ti67584b98.
- Seed et al. (2025) ByteDance Seed, Yufeng Yuan, Yu Yue, Mingxuan Wang, Xiaochen Zuo, Jiaze Chen, Lin Yan, Wenyuan Xu, Chi Zhang, Xin Liu, et al. Seed-thinking-v1. 5: Advancing superb reasoning models with reinforcement learning. arXiv preprint arXiv:2504.13914, 2025.
- Shao et al. (2024) Zhihong Shao, Peiyi Wang, Qihao Zhu, Runxin Xu, Junxiao Song, Xiao Bi, Haowei Zhang, Mingchuan Zhang, YK Li, Y Wu, et al. Deepseekmath: Pushing the limits of mathematical reasoning in open language models. arXiv preprint arXiv:2402.03300, 2024.
- Suzgun et al. (2022) Mirac Suzgun, Nathan Scales, Nathanael Schärli, Sebastian Gehrmann, Yi Tay, Hyung Won Chung, Aakanksha Chowdhery, Quoc V Le, Ed H Chi, Denny Zhou, et al. Challenging big-bench tasks and whether chain-of-thought can solve them. arXiv preprint arXiv:2210.09261, 2022.
- Xie et al. (2025a) Chulin Xie, Yangsibo Huang, Chiyuan Zhang, Da Yu, Xinyun Chen, Bill Yuchen Lin, Bo Li, Badih Ghazi, and Ravi Kumar. On memorization of large language models in logical reasoning, 2025a. URL https://arxiv.org/abs/2410.23123.
- Xie et al. (2025b) Tian Xie, Zitian Gao, Qingnan Ren, Haoming Luo, Yuqian Hong, Bryan Dai, Joey Zhou, Kai Qiu, Zhirong Wu, and Chong Luo. Logic-rl: Unleashing llm reasoning with rule-based reinforcement learning. arXiv preprint arXiv:2502.14768, 2025b.
- Yang et al. (2024) An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, et al. Qwen2. 5 technical report. arXiv preprint arXiv:2412.15115, 2024.
- Yu et al. (2025) Qiying Yu, Zheng Zhang, Ruofei Zhu, Yufeng Yuan, Xiaochen Zuo, Yu Yue, Tiantian Fan, Gaohong Liu, Lingjun Liu, Xin Liu, et al. Dapo: An open-source llm reinforcement learning system at scale. arXiv preprint arXiv:2503.14476, 2025.
- Zeng et al. (2025a) Huaye Zeng, Dongfu Jiang, Haozhe Wang, Ping Nie, Xiaotong Chen, and Wenhu Chen. Acecoder: Acing coder rl via automated test-case synthesis. ArXiv, 2502.01718, 2025a.
- Zeng et al. (2025b) Weihao Zeng, Yuzhen Huang, Qian Liu, Wei Liu, Keqing He, Zejun Ma, and Junxian He. Simplerl-zoo: Investigating and taming zero reinforcement learning for open base models in the wild. arXiv preprint arXiv:2503.18892, 2025b.
- Zhang et al. (2025) Xiaojiang Zhang, Jinghui Wang, Zifei Cheng, Wenhao Zhuang, Zheng Lin, Minglei Zhang, Shaojie Wang, Yinghan Cui, Chao Wang, Junyi Peng, et al. Srpo: A cross-domain implementation of large-scale reinforcement learning on llm. arXiv preprint arXiv:2504.14286, 2025.
Appendix A Comprehensive Overview of SynLogic
A.1 Task Composition and Sources
Table 4 presents the diverse collection of tasks incorporated in SynLogic. We have carefully selected these tasks from established benchmarks including KOR-Bench (Ma et al., 2024), BBH (Suzgun et al., 2022), and BBEH (Kazemi et al., 2025). Additionally, we integrated some logical reasoning tasks not previously featured in these benchmarks, such as Mathador (Kurtic et al., 2024) and Minesweeper (Li et al., 2024).
Our collection comprises 35 distinct tasks, with only two (Zebra Puzzle (Lin et al., 2024) and ARC-AGI (Chollet, 2019)) using existing data sources. For all remaining tasks, we generated custom datasets. Importantly, we developed and implemented verifiers for all tasks in the collection, ensuring consistent evaluation across the benchmark.
No. | Task Name | Description |
1 | ARC-AGI | A collection of general intelligence tasks requiring abstract reasoning and pattern recognition. |
2 | Arrow Maze | A maze-solving task where arrows dictate movement, requiring pathfinding logic. |
3 | Boolean Expressions | Evaluating logical expressions with AND, OR, NOT operators. |
4 | Buggy Tables | Correcting flawed data in tables based on logical constraints. |
5 | Calcudoko | A math-based Sudoku variant with arithmetic constraints. |
6 | Campsite | Placing tents in a grid while satisfying adjacency rules. |
7 | Cipher | Decoding encrypted messages based on given rules. |
8 | Cryptarithm | Solving arithmetic puzzles where letters represent digits. |
9 | Dyck Language | Validating bracket sequences for correct nesting. |
10 | Dyck Language Errors | Identifying and correcting errors in bracket sequences. |
11 | Dyck Language Reasoning Errors | Advanced reasoning for errors in bracket nesting logic. |
12 | Futoshiki | A grid-based logic puzzle with inequality constraints. |
13 | Goods Exchange | Tracking item exchanges among multiple participants. |
14 | Kukurasu | A grid-based puzzle involving row/column weight sums. |
15 | Mathador | A math strategy game involving arithmetic operations. |
16 | Math Path | Finding correct numbers to satisfy equations in a grid. |
17 | Minesweeper | Logical deduction to uncover mines on a grid. |
18 | Norinori | Placing domino tiles in a grid with adjacency rules. |
19 | Number Wall | Constructing walls to separate grid regions based on rules. |
20 | Numbrix | Filling a grid with consecutive numbers in order. |
21 | Object Counting | Counting specific objects under given constraints. |
22 | Object Properties | Inferring and reasoning about object attributes. |
23 | Operation | Solving puzzles with custom-defined mathematical operations. |
24 | Skyscraper Puzzle | Determining building heights based on visibility clues. |
25 | Space Reasoning | Reasoning about spatial relationships in a grid. |
26 | Space Reasoning Tree | Advanced spatial reasoning tasks with hierarchical relationships. |
27 | Star Placement Puzzle | Placing stars in a grid while avoiding adjacency conflicts. |
28 | Sudoku | Solving the classic number-placement puzzle. |
29 | Survo | Filling grids to satisfy row and column sum constraints. |
30 | Time Sequence | Scheduling tasks or events with overlapping constraints. |
31 | Web of Lies | Determining truth-tellers and liars through logical statements. |
32 | Word Sorting | Sorting words based on custom rules or constraints. |
33 | Word Sorting Mistake | Identifying mistakes in word sorting logic or reasoning. |
34 | Wordscapes | A crossword puzzle where players fill words from lists while matching intersections. |
35 | Zebra Puzzle | Solving complex logic puzzles with multiple constraints. |
A.2 SynLogic-Hard and SynLogic-Easy
The SynLogic-Hard dataset encompasses all 35 tasks, representing a challenging upper bound calibrated to the solvability thresholds of DeepSeek R1 and OpenAI-o3-mini. During our experiments with Qwen2.5-32B-Base, we observed consistent training accuracy gains across this comprehensive task set. However, this difficulty level proved excessive for smaller-scale models like Qwen2.5-7B-Base. Despite reducing the difficulty parameters across tasks, eight specific tasks (Arrow Maze, Goods Exchange, Kukurasu, Minesweeper, Norinori, Object Counting, Space Reasoning Tree, and Wordscapes) persistently yielded zero accuracy when training the 7B models. Consequently, we developed SynLogic-Easy, a modified variant that excludes these particularly challenging tasks, to provide a more appropriate training dataset for the Qwen2.5-7B-Base model.
Appendix B Training and Evaluation Details
B.1 Training
B.1.1 Training Template
B.1.2 Training hyper-parameters
For 7B model training, we adopted the following hyper-parameters: with learning rate 1e-6, GRPO group size 16, max prompt length 2048, max response length 16384, prompt batch size 128, mini batch size 64, with clip high 0.28, clip low 0.2.
For 32B model training in §3, we adopted the following hyper-parameters: with learning rate 2e-6, GRPO group size 16, max prompt length 2048, max response length 28672, prompt batch size 128, mini batch size 16, with clip high 0.28, clip low 0.2.
For 32B model training in §4, we adopted the following hyper-parameters: with learning rate 2e-6, GRPO group size 16, max prompt length 2048, max response length 12288, prompt batch size 512, mini batch size 16, with clip high 0.28, clip low 0.2.
B.1.3 Training Dynamics Analysis
We present the training accuracy progression of our 7B and 32B models on the SynLogic dataset in Figure 9. The results demonstrate a consistent improvement in accuracy throughout the training process for both model sizes.


B.2 Evaluation



B.2.1 Performance Analysis of Mixed Training with Math
We analyze the training dynamics when combining our logical reasoning dataset with mathematical content. Figure 10 presents performance results across three mathematical benchmarks: MATH 500, AIME 2024, and AMC 2023. These results demonstrate that mixed training can achieve similar performance with the same number of training steps while using less mathematical data. This suggests that mixed training creates more efficient training dynamics.
B.3 Performance Analysis of Mixed Training with Coding
We also provide the training dynamics when mixed training with coding data in Figure 11. The observation is similar to mixed training with math, suggesting that mixed training leads to more efficient training dynamics.

